Phonetic and speaker variations in automatic emotion classification

نویسندگان

  • Vidhyasaharan Sethu
  • Eliathamby Ambikairajah
  • Julien Epps
چکیده

The speech signal contains information that characterises the speaker and the phonetic content, together with the emotion being expressed. This paper looks at the effect of this speakerand phoneme-specific information on speech-based automatic emotion classification. The performances of a classification system using established acoustic and prosodic features for different phonemes are compared, in both speaker-dependent and speaker-independent modes, using the LDC Emotional Prosody speech corpus. Results from these evaluations indicate that speaker variability is more significant than phonetic variations. They also suggest that some phonemes are easier to classify than others.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Improving Automatic Emotion Recognition from speech using Rhythm and Temporal feature

This paper is devoted to improve automatic emotion recognition from speech by incorporating rhythm and temporal features. Research on automatic emotion recognition so far has mostly been based on applying features like MFCC’s, pitch and energy/intensity. The idea focuses on borrowing rhythm features from linguistic and phonetic analysis and applying them to the speech signal on the basis of aco...

متن کامل

Asc12. Effects of Emotion on Different Phoneme Classes

This study investigates the effects of emotion on different phoneme classes using short-term spectral features. In the research on emotion in speech, most studies have focused on prosodic features of speech. In this study, based on the hypothesis that different emotions have varying effects on the properties of the different speech sounds, we investigate the usefulness of phoneme-class level ac...

متن کامل

A Comparative Study of Gender and Age Classification in Speech Signals

Accurate gender classification is useful in speech and speaker recognition as well as speech emotion classification, because a better performance has been reported when separate acoustic models are employed for males and females. Gender classification is also apparent in face recognition, video summarization, human-robot interaction, etc. Although gender classification is rather mature in a...

متن کامل

Relation between geometry and kinematics of articulatory trajectory associated with emotional speech production

We investigate whether articulatory movement trajectories follow the nonlinear invariant relationship between the tangential velocity and the curvature, i.e. the one-third power law. The power law holds for articulatory trajectories of phonetic rendering, although the exponent is in the range 0.35–0.42 when averaged across speakers and emotions but is relatively invariant under speaker/emotion ...

متن کامل

Automatic Derivation of Phonetic Rules by an Iterated Normalisation Procedure

In this paper an iterative normalisation procedure to automatically derive phonetic rules from a labelled speech corpus is described. It is assumed that the acoustic influence of coarticulatory constraints can be superimposed to model natural spectral variation. The algorithm proves to be promising when used to analyse the effect of phonetic context, stress and duration of Swedish front vowels ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2008